Make local variables out of nn.sequential members for onnx exportability #14
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
The pytorch onnx exporter doesn't work if parameters are included as members of two modules. This means the pattern of having self.operator and passing self.operator to nn.sequential breaks the ability to export an onnx model.
This updates the TCN network to mirror the implementation of many other networks in pytorch where the operators passed to nn.Sequential are all local variables, and weight initialization is done at construction time to avoid the need for member variables for everything.